-
-
Notifications
You must be signed in to change notification settings - Fork 5.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[V1] Improve AsyncLLM
and API Server
#11237
[V1] Improve AsyncLLM
and API Server
#11237
Conversation
👋 Hi! Thank you for contributing to the vLLM project. Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can do one of these:
🚀 |
This pull request has merge conflicts that must be resolved before it can be |
for out in outputs: | ||
# Note: it is possible that a request was aborted | ||
# due to client cancellation while EngineCoreOutputs | ||
# are still flowing, so we just ignore. | ||
if out.request_id in self.rid_to_queue: | ||
self.rid_to_queue[out.request_id].put_nowait(out) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do you think this isn't worth logging since we are sure the request was aborted?
Co-authored-by: Michael Goin <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
With the http middleware issue fixed - curious how much perf diff the new 3 process architecture yields?
This pull request has merge conflicts that must be resolved before it can be |
Closing as complete |
NOTE: I am in process of slipping this PR into separate pieces
SUMMARY:
add_request_id
middleware optionalAsyncLLM
performance with 3 process architecture and reduction of task switching inAsyncLLM
SIGQUIT
mp.Pipe
ulimit
TODO: